2023-12-12 10:15:16.AIbase.4.1k
Research on the Risks of Training Data Leakage in ChatGPT
The research found that large language models like ChatGPT retain a considerable amount of original training data during their training process. Attackers can extract large volumes of training data from the model using specific attack methods, threatening the privacy of data owners. Researchers utilized a concept of 'extractable memory' attack techniques and discovered that existing extraction attacks can recover a significant amount of training data from memory even without using the actual training data as prompts. ⚠️ Researchers recommend that data security protective measures should be implemented while developing and using large language models.